break;
case HVMOP_flush_tlbs:
- rc = guest_handle_is_null(arg) ? hvmop_flush_tlb_all() : -ENOSYS;
+ rc = guest_handle_is_null(arg) ? hvmop_flush_tlb_all() : -EINVAL;
break;
case HVMOP_track_dirty_vram:
{
case XEN_DOMCTL_DEBUG_OP_SINGLE_STEP_ON:
case XEN_DOMCTL_DEBUG_OP_SINGLE_STEP_OFF:
- rc = -ENOSYS;
+ rc = -EOPNOTSUPP;
if ( !cpu_has_monitor_trap_flag )
break;
rc = 0;
struct get_reserved_device_memory grdm;
if ( unlikely(start_extent) )
- return -ENOSYS;
+ return -EINVAL;
if ( copy_from_guest(&grdm.map, compat, 1) ||
!compat_handle_okay(grdm.map.buffer, grdm.map.nr_entries) )
case XENMEM_exchange:
if ( unlikely(start_extent) )
- return -ENOSYS;
+ return -EINVAL;
rc = memory_exchange(guest_handle_cast(arg, xen_memory_exchange_t));
break;
case XENMEM_maximum_ram_page:
if ( unlikely(start_extent) )
- return -ENOSYS;
+ return -EINVAL;
rc = max_page;
break;
case XENMEM_maximum_reservation:
case XENMEM_maximum_gpfn:
if ( unlikely(start_extent) )
- return -ENOSYS;
+ return -EINVAL;
if ( copy_from_guest(&domid, arg, 1) )
return -EFAULT;
struct page_info *page;
if ( unlikely(start_extent) )
- return -ENOSYS;
+ return -EINVAL;
if ( copy_from_guest(&xrfp, arg, 1) )
return -EFAULT;
case XENMEM_claim_pages:
if ( unlikely(start_extent) )
- return -ENOSYS;
+ return -EINVAL;
if ( copy_from_guest(&reservation, arg, 1) )
return -EFAULT;
struct vnuma_info tmp;
if ( unlikely(start_extent) )
- return -ENOSYS;
+ return -EINVAL;
/*
* Guest passes nr_vnodes, number of regions and nr_vcpus thus
struct get_reserved_device_memory grdm;
if ( unlikely(start_extent) )
- return -ENOSYS;
+ return -EINVAL;
if ( copy_from_guest(&grdm.map, arg, 1) ||
!guest_handle_okay(grdm.map.buffer, grdm.map.nr_entries) )